# Sliding Window Attention
H2o Danube 1.8b Base
Apache-2.0
A 1.8B parameter base language model trained by H2O.ai, based on an improved Llama 2 architecture with 16K context length support
Large Language Model
Transformers English

H
h2oai
281
43
Mistral 7B Instruct V0.1
Apache-2.0
Mistral-7B-Instruct-v0.1 is a version of the Mistral-7B-v0.1 generative text model fine-tuned with various public dialogue datasets for instruction following.
Large Language Model
Transformers

M
mistralai
468.63k
1,659
Nat Base In1k 224
MIT
NAT-Base is a Vision Transformer model trained on ImageNet-1K, which uses the neighborhood attention mechanism for image classification.
Image Classification
Transformers Other

N
shi-labs
6
0
Nat Small In1k 224
MIT
NAT-Small is a hierarchical vision transformer based on neighborhood attention, designed for image classification tasks.
Image Classification
Transformers Other

N
shi-labs
6
0
Nat Mini In1k 224
MIT
NAT-Mini is a lightweight vision Transformer model based on neighborhood attention mechanism, designed for ImageNet image classification tasks
Image Classification
Transformers Other

N
shi-labs
109
0
Dinat Mini In1k 224
MIT
DiNAT-Mini is a hierarchical vision Transformer model based on neighborhood attention mechanism, specifically designed for image classification tasks.
Image Classification
Transformers

D
shi-labs
462
1
Swin Tiny Patch4 Window7 224 Finetuned Braintumordata
Apache-2.0
Vision model based on Swin Transformer architecture, fine-tuned specifically for brain tumor image analysis
Image Classification
Transformers

S
surajjoshi
11
1
Longformer Base 4096 Spanish
MIT
A Spanish long-document processing model developed based on RoBERTa checkpoints, supporting sequence lengths of up to 4096 tokens
Large Language Model
Transformers Spanish

L
mrm8488
22
16
Featured Recommended AI Models